An Efficient, Sparsity-Preserving, Online Algorithm for Data Approximation
نویسندگان
چکیده
The Singular Value Decomposition (SVD) is a longstanding standard for data approximation because it is optimal in the 2 and Frobenius norms. The SVD, nevertheless, suffers from many setbacks, including computational cost, loss of sparsity in the decomposition, and the inability to be updated easily when new information arrives. Additionally, the SVD provides limited information on data features and variables that best represent the data. In this work, we present a truncated LU factorization called Spectrum-Revealing LU (SRLU) for effective low-rank matrix approximation, and develop the first algorithm to compute an SRLU factorization, which is both efficient and reliable. Our algorithm uses randomization and a novel LU updating technique with partial pivoting, which is more stable than any other known LU updating algorithm. We provide both approximation error bounds and singular value bounds for the SRLU approximation computed by our algorithm. Our analysis suggests that SRLU is competitive with the best low-rank matrix approximation methods, deterministic or randomized, in both computational complexity and approximation quality. Numeric experiments illustrate that SRLU preserves sparsity, highlights important data features and variables, can be efficiently updated, and calculates data approximations nearly as accurately as the SVD. To the best of our knowledge this is the first practical variant of the LU decomposition for efficient and effective low-rank matrix approximation.
منابع مشابه
An Efficient, Sparsity-Preserving Online Algorithm for Data Approximation: Supplementary Material
001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050 051 052 053 054 055 056 057 058 059 060 061 062 063 064 065 066 067 068 069 070 071 072 073 074 075 076 077 078 079 080 081 082 083 084 085 086 087 088 089 090 091 092 093 094 095 096 097 098 099 1...
متن کاملAn Efficient, Sparsity-Preserving, Online Algorithm for Low-Rank Approximation
Low-rank matrix approximation is a fundamental tool in data analysis for processing large datasets, reducing noise, and finding important signals. In this work, we present a novel truncated LU factorization called Spectrum-Revealing LU (SRLU) for effective low-rank matrix approximation, and develop a fast algorithm to compute an SRLU factorization. We provide both matrix and singular value appr...
متن کاملSemi-supervised Sparsity Pairwise Constraint Preserving Projections based on GA
The deficiency of the ability for preserving global geometric structure information of data is the main problem of existing semi-supervised dimensionality reduction with pairwise constraints. A dimensionality reduction algorithm called Semi-supervised Sparsity Pairwise Constraint Preserving Projections based on Genetic Algorithm (SSPCPPGA) is proposed. On the one hand, the algorithm fuses unsup...
متن کاملEfficient Approximation Algorithms for Point-set Diameter in Higher Dimensions
We study the problem of computing the diameter of a set of $n$ points in $d$-dimensional Euclidean space for a fixed dimension $d$, and propose a new $(1+varepsilon)$-approximation algorithm with $O(n+ 1/varepsilon^{d-1})$ time and $O(n)$ space, where $0 < varepsilonleqslant 1$. We also show that the proposed algorithm can be modified to a $(1+O(varepsilon))$-approximation algorithm with $O(n+...
متن کاملA centralized privacy-preserving framework for online social networks
There are some critical privacy concerns in the current online social networks (OSNs). Users' information is disclosed to different entities that they were not supposed to access. Furthermore, the notion of friendship is inadequate in OSNs since the degree of social relationships between users dynamically changes over the time. Additionally, users may define similar privacy settings for their f...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1602.05950 شماره
صفحات -
تاریخ انتشار 2016